Neural mixture models with expectation-maximization for end-to-end deep clustering
نویسندگان
چکیده
Any clustering algorithm must synchronously learn to model the clusters and allocate data those in absence of labels. Mixture model-based methods with pre-defined statistical distributions based on cluster likelihoods. They iteratively refine distribution parameters member assignments following Expectation-Maximization (EM) algorithm. However, representability such hand-designed that employ a limited amount is not adequate for most real-world tasks. In this paper, we realize mixture neural network where final layer neurons, aid an additional transformation, approximate outputs. The pose as distributions. result elegant, much-generalized representation than restricted We train end-to-end via batch-wise EM iterations forward pass acts E-step backward M-step. image clustering, mixture-based objective can be used along existing learning methods. particular, show when mixture-EM optimization fused consistency optimization, it improves sole performance clustering. Our trained networks outperform single-stage deep still depend k-means, unsupervised classification accuracy 63.8% STL10, 58% CIFAR10, 25.9% CIFAR100, 98.9% MNIST.
منابع مشابه
Mixture Models and Expectation-Maximization
This tutorial attempts to provide a gentle introduction to EM by way of simple examples involving maximum-likelihood estimation of mixture-model parameters. Readers familiar with ML paramter estimation and clustering may want to skip directly to Sections 5.2 and 5.3.
متن کاملSupervised Hashing with End-to-End Binary Deep Neural Network
Image hashing is a popular technique applied to large scale content-based visual retrieval due to its compact and efficient binary codes. Our work proposes a new end-to-end deep network architecture for supervised hashing which directly learns binary codes from input images and maintains good properties over binary codes such as similarity preservation, independence, and balancing. Furthermore,...
متن کاملEnd-to-End Optimized Speech Coding with Deep Neural Networks
Modern compression algorithms are often the result of laborious domain-specific research; industry standards such as MP3, JPEG, and AMR-WB took years to develop and were largely hand-designed. We present a deep neural network model which optimizes all the steps of a wideband speech coding pipeline (compression, quantization, entropy coding, and decompression) end-to-end directly from raw speech...
متن کاملEnd-to-End Deep Neural Network for Automatic Speech Recognition
We investigate the efficacy of deep neural networks on speech recognition. Specifically, we implement an end-to-end deep learning system that utilizes mel-filter bank features to directly output to spoken phonemes without the need of a traditional Hidden Markov Model for decoding. The system will comprise of two variants of neural networks for phoneme recognition. In particular, we utilize conv...
متن کاملDeep Speaker: an End-to-End Neural Speaker Embedding System
We present Deep Speaker, a neural speaker embedding system that maps utterances to a hypersphere where speaker similarity is measured by cosine similarity. The embeddings generated by Deep Speaker can be used for many tasks, including speaker identification, verification, and clustering. We experiment with ResCNN and GRU architectures to extract the acoustic features, then mean pool to produce ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Neurocomputing
سال: 2022
ISSN: ['0925-2312', '1872-8286']
DOI: https://doi.org/10.1016/j.neucom.2022.07.017